-
Notifications
You must be signed in to change notification settings - Fork 2.5k
Add new kimi model to groq #7692
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thank you for your contribution! I've reviewed the changes and found several issues that need attention before this can be merged.
| | "openai/gpt-oss-20b" | ||
|
|
||
| export const groqDefaultModelId: GroqModelId = "llama-3.3-70b-versatile" // Defaulting to Llama3 70B Versatile | ||
| export const groqDefaultModelId: GroqModelId = "moonshotai/kimi-k2-instruct-0905" |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
llama-3.3-70b-versatile to moonshotai/kimi-k2-instruct-0905 will affect all users who rely on the default model. They may experience different behavior, costs, and capabilities. Could we consider:
- Keeping the existing default for backward compatibility
- Or documenting this breaking change prominently in the changelog/release notes?
| contextWindow: 262144, | ||
| supportsImages: false, | ||
| supportsPromptCache: true, | ||
| inputPrice: 0.6, |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Is the significant price reduction intentional? The new model has:
- Input price: /bin/sh.6 (40% lower than the existing kimi-k2-instruct at .0)
- Output price: .5 (17% lower than the existing kimi-k2-instruct at .0)
Could you confirm these prices are correct according to Groq's pricing?
| supportsPromptCache: true, | ||
| inputPrice: 0.6, | ||
| outputPrice: 2.5, | ||
| cacheReadsPrice: 0.15, |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
The cache reads price of /bin/sh.15 represents a 75% discount from the input price (/bin/sh.6), while the existing kimi-k2-instruct model has a 50% discount. Is this aggressive discount structure correct?
| inputPrice: 0.6, | ||
| outputPrice: 2.5, | ||
| cacheReadsPrice: 0.15, | ||
| description: "Moonshot AI Kimi K2 Instruct 1T model, 256K context.", |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Minor inconsistency: The description says "256K context" but the contextWindow value is 262144 (which is 256Ki or 256 * 1024). While technically correct, this differs from the pattern used in the existing model which says "128K context" with value 131072. Should we use "256Ki context" or keep it consistent with the existing pattern?
Important
Add
moonshotai/kimi-k2-instruct-0905model togroq.tsand set it as the default model with specific configurations.moonshotai/kimi-k2-instruct-0905toGroqModelIdandgroqModelsingroq.ts.groqDefaultModelIdtomoonshotai/kimi-k2-instruct-0905ingroq.ts.moonshotai/kimi-k2-instruct-0905hasmaxTokens: 16384,contextWindow: 262144,supportsImages: false,supportsPromptCache: true,inputPrice: 0.6,outputPrice: 2.5,cacheReadsPrice: 0.15.This description was created by
for c43f39e. You can customize this summary. It will automatically update as commits are pushed.